找回密码
 To register

QQ登录

只需一步,快速开始

扫一扫,访问微社区

Titlebook: Deep Learning for Video Understanding; Zuxuan Wu,Yu-Gang Jiang Book 2024 The Editor(s) (if applicable) and The Author(s), under exclusive

[复制链接]
查看: 12318|回复: 45
发表于 2025-3-21 20:09:09 | 显示全部楼层 |阅读模式
书目名称Deep Learning for Video Understanding
编辑Zuxuan Wu,Yu-Gang Jiang
视频video
概述Presents an overview of deep learning techniques for video understanding.Covers important topics like action recognition, action localization, video captioning, and more.Introduces cutting-edge and st
丛书名称Wireless Networks
图书封面Titlebook: Deep Learning for Video Understanding;  Zuxuan Wu,Yu-Gang Jiang Book 2024 The Editor(s) (if applicable) and The Author(s), under exclusive
描述.This book presents deep learning techniques for video understanding. For deep learning basics, the authors cover machine learning pipelines and notations, 2D and 3D Convolutional Neural Networks for spatial and temporal feature learning. For action recognition, the authors introduce classical frameworks for image classification, and then elaborate both image-based and clip-based 2D/3D CNN networks for action recognition. For action detection, the authors elaborate sliding windows, proposal-based detection methods, single stage and two stage approaches, spatial and temporal action localization, followed by datasets introduction. For video captioning, the authors present language-based models and how to perform sequence to sequence learning for video captioning. For unsupervised feature learning, the authors discuss the necessity of shifting from supervised learning to unsupervised learning and then introduce how to design better surrogate training tasks to learn video representations. Finally, the book introduces recent self-training pipelines like contrastive learning and masked image/video modeling with transformers. The book provides promising directions, with an aim to promote
出版日期Book 2024
关键词action recognition; video captioning; action localization; motion extraction; spatial-temporal feature l
版次1
doihttps://doi.org/10.1007/978-3-031-57679-9
isbn_softcover978-3-031-57681-2
isbn_ebook978-3-031-57679-9Series ISSN 2366-1186 Series E-ISSN 2366-1445
issn_series 2366-1186
copyrightThe Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Switzerl
The information of publication is updating

书目名称Deep Learning for Video Understanding影响因子(影响力)




书目名称Deep Learning for Video Understanding影响因子(影响力)学科排名




书目名称Deep Learning for Video Understanding网络公开度




书目名称Deep Learning for Video Understanding网络公开度学科排名




书目名称Deep Learning for Video Understanding被引频次




书目名称Deep Learning for Video Understanding被引频次学科排名




书目名称Deep Learning for Video Understanding年度引用




书目名称Deep Learning for Video Understanding年度引用学科排名




书目名称Deep Learning for Video Understanding读者反馈




书目名称Deep Learning for Video Understanding读者反馈学科排名




单选投票, 共有 0 人参与投票
 

0票 0%

Perfect with Aesthetics

 

0票 0%

Better Implies Difficulty

 

0票 0%

Good and Satisfactory

 

0票 0%

Adverse Performance

 

0票 0%

Disdainful Garbage

您所在的用户组没有投票权限
发表于 2025-3-21 21:14:08 | 显示全部楼层
Book 2024ng and then introduce how to design better surrogate training tasks to learn video representations. Finally, the book introduces recent self-training pipelines like contrastive learning and masked image/video modeling with transformers. The book provides promising directions, with an aim to promote
发表于 2025-3-22 01:57:04 | 显示全部楼层
the hotbeds of pretext tasks, which refer to network optimization tasks based on surrogate signals without human supervision, facilitating better performance on video-related downstream tasks. In this chapter, we undertake a comprehensive review of UVL, which begins with a preliminary introduction o
发表于 2025-3-22 08:19:14 | 显示全部楼层
发表于 2025-3-22 09:48:32 | 显示全部楼层
2366-1186 n, video captioning, and more.Introduces cutting-edge and st.This book presents deep learning techniques for video understanding. For deep learning basics, the authors cover machine learning pipelines and notations, 2D and 3D Convolutional Neural Networks for spatial and temporal feature learning. F
发表于 2025-3-22 15:10:33 | 显示全部楼层
发表于 2025-3-22 18:20:38 | 显示全部楼层
Angst – Bedingung des Mensch-Seinsirections, e.g., the construction of large-scale video foundation models, the application of large language models (LLMs) in video understanding, etc. By depicting these exciting prospects, we encourage the readers to embark on new endeavors to contribute to the advancement of this field.
发表于 2025-3-22 23:24:38 | 显示全部楼层
Book 2024ions, 2D and 3D Convolutional Neural Networks for spatial and temporal feature learning. For action recognition, the authors introduce classical frameworks for image classification, and then elaborate both image-based and clip-based 2D/3D CNN networks for action recognition. For action detection, th
发表于 2025-3-23 02:50:19 | 显示全部楼层
,I. Führung der eigenen Person,en successively proposed, promoting this large field to becoming more and more mature. In this chapter, we will briefly introduce the above aspects and travel through the corridors of time to systematically review the chronology of this dynamic field.
发表于 2025-3-23 09:34:22 | 显示全部楼层
Fallstudien „Führung von Experten“ons of these backbones. By the end of the chapter, readers will have a solid understanding of the basics of deep learning for video understanding and be well-equipped to explore more advanced topics in this exciting field.
 关于派博传思  派博传思旗下网站  友情链接
派博传思介绍 公司地理位置 论文服务流程 影响因子官网 SITEMAP 大讲堂 北京大学 Oxford Uni. Harvard Uni.
发展历史沿革 期刊点评 投稿经验总结 SCIENCEGARD IMPACTFACTOR 派博系数 清华大学 Yale Uni. Stanford Uni.
|Archiver|手机版|小黑屋| 派博传思国际 ( 京公网安备110108008328) GMT+8, 2025-5-9 21:32
Copyright © 2001-2015 派博传思   京公网安备110108008328 版权所有 All rights reserved
快速回复 返回顶部 返回列表